We broke AI guardrails down to six categories.
We curated datasets and models that demonstrate the state of AI safety using LLMs and other open source models.
| Developer | Model | Latency | Metric |
|---|---|---|---|
| Guardrails AI | Toxic Language | 0.0086 ms | 0.7182 |
| Natural Language Content Safety | 0.1063 ms | 0.5955 | |
| Microsoft | Azure Content Safety | 0.0632 ms | 0.5125 |
| Developer | Samples |
|---|---|
| toxic | 6090 |
| obscene | 3691 |
| insult | 3427 |
| identity_hate | 712 |
| severe_toxic | 367 |
| threat | 211 |